RLAttn: An actor-critic model of eye movements during category learning
نویسندگان
چکیده
Here we introduce a simple actor-critic model of eye movements during category learning that we call RLAttn (Reinforcement Learning of Attention). RLAttn stores the rewards it receives for making decisions or performing actions, while attempting to associate stimuli with particular categories. Over multiple trials, RLAttn learns that a large reward is most likely when the values of the relevant stimulus features have been revealed by fixations to them. The model is able to approximate human learning curves in a common category structure while generating fixation patterns similar to those found in human eye tracking data. We additionally observed that the model reduces its fixation counts to irrelevant features over the course of learning. We conclude with a discussion on the effective role eye movements might play in bridging structural credit assignment and temporal credit assignment problems.
منابع مشابه
G Uide a Ctor - C Ritic for C Ontinuous C Ontrol
Actor-critic methods solve reinforcement learning problems by updating a parameterized policy known as an actor in a direction that increases an estimate of the expected return known as a critic. However, existing actor-critic methods only use values or gradients of the critic to update the policy parameter. In this paper, we propose a novel actor-critic method called the guide actor-critic (GA...
متن کاملAn Actor-critic Algorithm for Learning Rate Learning
Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks. It is observed that the models trained by SGD are sensitive to learning rates and good learning rates are problem specific. To avoid manually searching of learning rates, whic...
متن کاملActor-Critic Control with Reference Model Learning
We propose a new actor-critic algorithm for reinforcement learning. The algorithm does not use an explicit actor, but learns a reference model which represents a desired behaviour, along which the process is to be controlled by using the inverse of a learned process model. The algorithm uses Local Linear Regression (LLR) to learn approximations of all the functions involved. The online learning...
متن کاملReinforcement Learning for Learning Rate Control
Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks. It is observed that the models trained by SGD are sensitive to learning rates and good learning rates are problem specific. We propose an algorithm to automatically learn lear...
متن کاملOnActor-Critic Algorithms
In this article, we propose and analyze a class of actor-critic algorithms. These are two-time-scale algorithms in which the critic uses temporal difference learning with a linearly parameterized approximation architecture, and the actor is updated in an approximate gradient direction, based on information provided by the critic. We show that the features for the critic should ideally span a su...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014